72 research outputs found

    Spike Onset Dynamics and Response Speed in Neuronal Populations

    Full text link
    Recent studies of cortical neurons driven by fluctuating currents revealed cutoff frequencies for action potential encoding of several hundred Hz. Theoretical studies of biophysical neuron models have predicted a much lower cutoff frequency of the order of average firing rate or the inverse membrane time constant. The biophysical origin of the observed high cutoff frequencies is thus not well understood. Here we introduce a neuron model with dynamical action potential generation, in which the linear response can be analytically calculated for uncorrelated synaptic noise. We find that the cutoff frequencies increase to very large values when the time scale of action potential initiation becomes short

    Extracting non-linear integrate-and-fire models from experimental data using dynamic I–V curves

    Get PDF
    The dynamic I–V curve method was recently introduced for the efficient experimental generation of reduced neuron models. The method extracts the response properties of a neuron while it is subject to a naturalistic stimulus that mimics in vivo-like fluctuating synaptic drive. The resulting history-dependent, transmembrane current is then projected onto a one-dimensional current–voltage relation that provides the basis for a tractable non-linear integrate-and-fire model. An attractive feature of the method is that it can be used in spike-triggered mode to quantify the distinct patterns of post-spike refractoriness seen in different classes of cortical neuron. The method is first illustrated using a conductance-based model and is then applied experimentally to generate reduced models of cortical layer-5 pyramidal cells and interneurons, in injected-current and injected- conductance protocols. The resulting low-dimensional neuron models—of the refractory exponential integrate-and-fire type—provide highly accurate predictions for spike-times. The method therefore provides a useful tool for the construction of tractable models and rapid experimental classification of cortical neurons

    Population density equations for stochastic processes with memory kernels

    Get PDF
    We present a method for solving population density equations (PDEs)–-a mean-field technique describing homogeneous populations of uncoupled neurons—where the populations can be subject to non-Markov noise for arbitrary distributions of jump sizes. The method combines recent developments in two different disciplines that traditionally have had limited interaction: computational neuroscience and the theory of random networks. The method uses a geometric binning scheme, based on the method of characteristics, to capture the deterministic neurodynamics of the population, separating the deterministic and stochastic process cleanly. We can independently vary the choice of the deterministic model and the model for the stochastic process, leading to a highly modular numerical solution strategy. We demonstrate this by replacing the master equation implicit in many formulations of the PDE formalism by a generalization called the generalized Montroll-Weiss equation—a recent result from random network theory—describing a random walker subject to transitions realized by a non-Markovian process. We demonstrate the method for leaky- and quadratic-integrate and fire neurons subject to spike trains with Poisson and gamma-distributed interspike intervals. We are able to model jump responses for both models accurately to both excitatory and inhibitory input under the assumption that all inputs are generated by one renewal process

    Motif Statistics and Spike Correlations in Neuronal Networks

    Get PDF
    Motifs are patterns of subgraphs of complex networks. We studied the impact of such patterns of connectivity on the level of correlated, or synchronized, spiking activity among pairs of cells in a recurrent network model of integrate and fire neurons. For a range of network architectures, we find that the pairwise correlation coefficients, averaged across the network, can be closely approximated using only three statistics of network connectivity. These are the overall network connection probability and the frequencies of two second-order motifs: diverging motifs, in which one cell provides input to two others, and chain motifs, in which two cells are connected via a third intermediary cell. Specifically, the prevalence of diverging and chain motifs tends to increase correlation. Our method is based on linear response theory, which enables us to express spiking statistics using linear algebra, and a resumming technique, which extrapolates from second order motifs to predict the overall effect of coupling on network correlation. Our motif-based results seek to isolate the effect of network architecture perturbatively from a known network state

    Stabilisation of beta and gamma oscillation frequency in the mammalian olfactory bulb

    Get PDF
    International audienceThe dynamics of the mammalian olfactory bulb (OB) is characterized by local field potential (LFP) oscillations either slow, in the theta range (2-10Hz, tightly linked to the respiratory rhythm), or fast, in the beta (15-30Hz) or gamma (40-90Hz) range. These fast oscillations are known to be modulated by odorant features and animal experience or state, but both their mechanisms and implication in coding are still not well understood. In this study, we used a double canulation protocol to impose artificial breathing rhythms to anesthetized rats while recording the LFP in the OB. We observed that despite the changes in the input air flow parameters (frequency or flow rate), the main characteristics of fast oscillations (duration, frequency or amplitude) were merely constant. We thus made the hypothesis that fast beta and gamma oscillations dynamics are entirely determined by the OB network properties and that external stimulation was only able put the network in a state which permits the generation of one or the other oscillations (they are never present simultaneously)

    A comparative study of different integrate-and-fire neurons: spontaneous activity, dynamical response, and stimulus-induced correlation

    Full text link
    Stochastic integrate-and-fire (IF) neuron models have found widespread applications in computational neuroscience. Here we present results on the white-noise-driven perfect, leaky, and quadratic IF models, focusing on the spectral statistics (power spectra, cross spectra, and coherence functions) in different dynamical regimes (noise-induced and tonic firing regimes with low or moderate noise). We make the models comparable by tuning parameters such that the mean value and the coefficient of variation of the interspike interval match for all of them. We find that, under these conditions, the power spectrum under white-noise stimulation is often very similar while the response characteristics, described by the cross spectrum between a fraction of the input noise and the output spike train, can differ drastically. We also investigate how the spike trains of two neurons of the same kind (e.g. two leaky IF neurons) correlate if they share a common noise input. We show that, depending on the dynamical regime, either two quadratic IF models or two leaky IFs are more strongly correlated. Our results suggest that, when choosing among simple IF models for network simulations, the details of the model have a strong effect on correlation and regularity of the output.Comment: 12 page

    Transient Responses to Rapid Changes in Mean and Variance in Spiking Models

    Get PDF
    The mean input and variance of the total synaptic input to a neuron can vary independently, suggesting two distinct information channels. Here we examine the impact of rapidly varying signals, delivered via these two information conduits, on the temporal dynamics of neuronal firing rate responses. We examine the responses of model neurons to step functions in either the mean or the variance of the input current. Our results show that the temporal dynamics governing response onset depends on the choice of model. Specifically, the existence of a hard threshold introduces an instantaneous component into the response onset of a leaky-integrate-and-fire model that is not present in other models studied here. Other response features, for example a decaying oscillatory approach to a new steady-state firing rate, appear to be more universal among neuronal models. The decay time constant of this approach is a power-law function of noise magnitude over a wide range of input parameters. Understanding how specific model properties underlie these response features is important for understanding how neurons will respond to rapidly varying signals, as the temporal dynamics of the response onset and response decay to new steady-state determine what range of signal frequencies a population of neurons can respond to and faithfully encode

    A Fokker-Planck formalism for diffusion with finite increments and absorbing boundaries

    Get PDF
    Gaussian white noise is frequently used to model fluctuations in physical systems. In Fokker-Planck theory, this leads to a vanishing probability density near the absorbing boundary of threshold models. Here we derive the boundary condition for the stationary density of a first-order stochastic differential equation for additive finite-grained Poisson noise and show that the response properties of threshold units are qualitatively altered. Applied to the integrate-and-fire neuron model, the response turns out to be instantaneous rather than exhibiting low-pass characteristics, highly non-linear, and asymmetric for excitation and inhibition. The novel mechanism is exhibited on the network level and is a generic property of pulse-coupled systems of threshold units.Comment: Consists of two parts: main article (3 figures) plus supplementary text (3 extra figures

    Numerical Solution of Differential Equations by the Parker-Sochacki Method

    Get PDF
    A tutorial is presented which demonstrates the theory and usage of the Parker-Sochacki method of numerically solving systems of differential equations. Solutions are demonstrated for the case of projectile motion in air, and for the classical Newtonian N-body problem with mutual gravitational attraction.Comment: Added in July 2010: This tutorial has been posted since 1998 on a university web site, but has now been cited and praised in one or more refereed journals. I am therefore submitting it to the Cornell arXiv so that it may be read in response to its citations. See "Spiking neural network simulation: numerical integration with the Parker-Sochacki method:" J. Comput Neurosci, Robert D. Stewart & Wyeth Bair and http://www.ncbi.nlm.nih.gov/pmc/articles/PMC2717378

    The location of the axon initial segment affects the bandwidth of spike initiation dynamics

    Get PDF
    The dynamics and the sharp onset of action potential (AP) generation have recently been the subject of intense experimental and theoretical investigations. According to the resistive coupling theory, an electrotonic interplay between the site of AP initiation in the axon and the somato-dendritic load determines the AP waveform. This phenomenon not only alters the shape of AP recorded at the soma, but also determines the dynamics of excitability across a variety of time scales. Supporting this statement, here we generalize a previous numerical study and extend it to the quantification of the input-output gain of the neuronal dynamical response. We consider three classes of multicompartmental mathematical models, ranging from ball-and-stick simplified descriptions of neuronal excitability to 3D-reconstructed biophysical models of excitatory neurons of rodent and human cortical tissue. For each model, we demonstrate that increasing the distance between the axonal site of AP initiation and the soma markedly increases the bandwidth of neuronal response properties. We finally consider the Liquid State Machine paradigm, exploring the impact of altering the site of AP initiation at the level of a neuronal population, and demonstrate that an optimal distance exists to boost the computational performance of the network in a simple classification task. Copyright
    corecore